Goto

Collaborating Authors

 perceptual similarity


Generating Images with Perceptual Similarity Metrics based on Deep Networks

Neural Information Processing Systems

We propose a class of loss functions, which we call deep perceptual similarity metrics (DeePSiM), allowing to generate sharp high resolution images from compressed abstract representations. Instead of computing distances in the image space, we compute distances between image features extracted by deep neural networks. This metric reflects perceptual similarity of images much better and, thus, leads to better results. We demonstrate two examples of use cases of the proposed loss: (1) networks that invert the AlexNet convolutional network; (2) a modified version of a variational autoencoder that generates realistic high-resolution random images.






0c0a7566915f4f24853fc4192689aa7e-AuthorFeedback.pdf

Neural Information Processing Systems

We thank the reviewers for their constructive comments on our paper. We address the major questions in the following. R1: The ability to handle large temporal inconsistency. In the experiments for most evaluated tasks, we do not observe the extreme cases mentioned by R1. The proposed IRT solves the multimodal inconsistency problem well, which is ignored by prior work.


DASH: A Meta-Attack Framework for Synthesizing Effective and Stealthy Adversarial Examples

Nafi, Abdullah Al Nomaan, Rahaman, Habibur, Haider, Zafaryab, Mahfuz, Tanzim, Suya, Fnu, Bhunia, Swarup, Chakraborty, Prabuddha

arXiv.org Artificial Intelligence

Numerous techniques have been proposed for generating adversarial examples in white-box settings under strict Lp-norm constraints. However, such norm-bounded examples often fail to align well with human perception, and only recently have a few methods begun specifically exploring perceptually aligned adversarial examples. Moreover, it remains unclear whether insights from Lp-constrained attacks can be effectively leveraged to improve perceptual efficacy. In this paper, we introduce DAASH, a fully differentiable meta-attack framework that generates effective and perceptually aligned adversarial examples by strategically composing existing Lp-based attack methods. DAASH operates in a multi-stage fashion: at each stage, it aggregates candidate adversarial examples from multiple base attacks using learned, adaptive weights and propagates the result to the next stage. A novel meta-loss function guides this process by jointly minimizing misclassification loss and perceptual distortion, enabling the framework to dynamically modulate the contribution of each base attack throughout the stages. We evaluate DAASH on adversarially trained models across CIFAR-10, CIFAR-100, and ImageNet. Despite relying solely on Lp-constrained based methods, DAASH significantly outperforms state-of-the-art perceptual attacks such as AdvAD -- achieving higher attack success rates (e.g., 20.63\% improvement) and superior visual quality, as measured by SSIM, LPIPS, and FID (improvements $\approx$ of 11, 0.015, and 5.7, respectively). Furthermore, DAASH generalizes well to unseen defenses, making it a practical and strong baseline for evaluating robustness without requiring handcrafted adaptive attacks for each new defense.